Optimal Control of Markov Processes
نویسنده
چکیده
The purpose of this article is to giye an overview of some recent developments in optimal stochastic control theory. The field has expanded a great deal during the last 20 years. It is not possible in this overview to go deeply into any topic, and a number of interesting topics have been omitted entirely. The list of references includes several books, conference proceedings and survey articles. The development of stochastic control theory has depended on parallel advances in the theory of stochastic processes and on certain topics iti partial differential equations. On the probabilistic side one can mention decomposition and representation theorems for semimartingales, formulas for absolutely continuous change of j)robability measure (e.g. the Girsanov formula), and the study of Ito-sense stochastic differential equations with discontinuous coefficients. It seems fair to say that these developments in stochastic processes were in turn to an extent influenced by their applications in stochastic control. For controlled Markov diffusion processes, there is a direct connection with certain nonlinear partial differential equations via the dynamic programming equation. These equations are of second order, elliptic or parabolic, and possibly degenerate. Stochastic control gives a way to represent their solutions probabilistically. There is an unforeseen connection with differential geometry via the Monge-Ampère equation. Broadly speaking, stochastic control theory deals with models of systems whose evolution is affected both by certain random influences
منابع مشابه
Utilizing Generalized Learning Automata for Finding Optimal Policies in MMDPs
Multi agent Markov decision processes (MMDPs), as the generalization of Markov decision processes to the multi agent case, have long been used for modeling multi agent system and are used as a suitable framework for Multi agent Reinforcement Learning. In this paper, a generalized learning automata based algorithm for finding optimal policies in MMDP is proposed. In the proposed algorithm, MMDP ...
متن کاملStochastic Dynamic Programming with Markov Chains for Optimal Sustainable Control of the Forest Sector with Continuous Cover Forestry
We present a stochastic dynamic programming approach with Markov chains for optimal control of the forest sector. The forest is managed via continuous cover forestry and the complete system is sustainable. Forest industry production, logistic solutions and harvest levels are optimized based on the sequentially revealed states of the markets. Adaptive full system optimization is necessary for co...
متن کاملCascade Markov Decision Processes: Theory and Applications
This paper considers the optimal control of time varying continuous time Markov chains whose transition rates are themselves Markov processes. In one set of problems the solution of an ordinary differential equation is shown to determine the optimal performance and feedback controls, while some other cases are shown to lead to singular optimal control problems which are more difficult to solve....
متن کاملOptimal Control of Markov Regenerative Processes
In the paper the integration of available results on SemiMarkov Decision Processes and on Markov Regenerative Processes is attempted, in order to de ne the mathematical framework for solving decision problems where the underlying structure state process is a Markov Regenerative Process, referred to as Markov Regenerative Decision Process. The essential question investigated here is which descri...
متن کاملMarkov Processes with Fixed Reciprocal
Continuing the work started in 11], in this paper we examine the construction of Gauss-Markov processes with xed reciprocal dynamics. We show how to construct Gauss-Markov processes, de-ned on a nite interval, having xed initial and end-point densities and belonging to a given reciprocal class. The problem of changing the end-point density of a Markov process, while remaining in the same recipr...
متن کاملOptimal Control of Partiality Observable Markov Processes over a Finite Horizon
This report presents an approach to find exact solution of optimal control of POMDPs (Partiality Observable Markov Decision Process) over a finite horizon under having a few reasonable assumptions. The approach only considers finite-state Markov processes. By comparing MDPs and PODMPs from optimal control policies point of view, it will be demonstrated that solving POMDPs is harder than solving...
متن کامل